Goto

Collaborating Authors

 unobserved confounder







Detecting Unobserved Confounders: A Kernelized Regression Approach

Chen, Yikai, Mao, Yunxin, Zheng, Chunyuan, Zou, Hao, Gu, Shanzhi, Liu, Shixuan, Shi, Yang, Yang, Wenjing, Kuang, Kun, Wang, Haotian

arXiv.org Machine Learning

Detecting unobserved confounders is crucial for reliable causal inference in observational studies. Existing methods require either linearity assumptions or multiple heterogeneous environments, limiting applicability to nonlinear single-environment settings. To bridge this gap, we propose Kernel Regression Confounder Detection (KRCD), a novel method for detecting unobserved confounding in nonlinear observational data under single-environment conditions. KRCD leverages reproducing kernel Hilbert spaces to model complex dependencies. By comparing standard and higherorder kernel regressions, we derive a test statistic whose significant deviation from zero indicates unobserved confounding. Theoretically, we prove two key results: First, in infinite samples, regression coefficients coincide if and only if no unobserved confounders exist. Second, finite-sample differences converge to zero-mean Gaussian distributions with tractable variance. Extensive experiments on synthetic benchmarks and the Twins dataset demonstrate that KRCD not only outperforms existing baselines but also achieves superior computational efficiency.


Policy Evaluation with Latent Confounders via Optimal Balance

Neural Information Processing Systems

Evaluating novel contextual bandit policies using logged data is crucial in applications where exploration is costly, such as medicine. But it usually relies on the assumption of no unobserved confounders, which is bound to fail in practice. We study the question of policy evaluation when we instead have proxies for the latent confounders and develop an importance weighting method that avoids fitting a latent outcome regression model. Surprisingly, we show that there exist no single set of weights that give unbiased evaluation regardless of outcome model, unlike the case with no unobserved confounders where density ratios are sufficient. Instead, we propose an adversarial objective and weights that minimize it, ensuring sufficient balance in the latent confounders regardless of outcome model. We develop theory characterizing the consistency of our method and tractable algorithms for it. Empirical results validate the power of our method when confounders are latent.




We completely agree that multi-decision confounding is also important, but

Neural Information Processing Systems

We'll address them all in the paper & answer the key questions here. We will add this discussion in the final version. D and relation to Z & B 2019 (R1) We briefly discussed this in Lines 76-80 and will highlight it further. We can minimize the loss using any model class (e.g. We only use the linearity assumption in our proof to guarantee statistical consistency.